Conceptual text region network: Cognition-inspired accurate scene text detection

نویسندگان

چکیده

Segmentation-based methods are widely used for scene text detection due to their superiority in describing arbitrary-shaped instances. However, two major problems still exist: (1) current label generation techniques mostly empirical and lack theoretical support, discouraging elaborate design; (2) as a result, most rely heavily on kernel segmentation which is unstable requires deliberate tuning. To address these challenges, we propose human cognition-inspired framework, termed, Conceptual Text Region Network (CTRNet). The framework utilizes Regions (CTRs), class of cognition-based tools inheriting good mathematical properties, allowing sophisticated design. Another component CTRNet an inference pipeline that, with the help CTRs, completely omits need segmentation. Compared previous segmentation-based methods, our approach not only more interpretable but also accurate. Experiment results show that achieves state-of-the-art performance benchmark CTW1500, Total-Text, MSRA-TD500, ICDAR 2015 datasets, yielding gains up 2.0%. Notably, best knowledge, among first models achieve F-measures higher than 85.0% all four benchmarks, demonstrating remarkable consistency stability.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Accurate Scene Text Recognition Based on Recurrent Neural Network

Scene text recognition is a useful but very challenging task due to uncontrolled condition of text in natural scenes. This paper presents a novel approach to recognize text in scene images. In the proposed technique, a word image is first converted into a sequential column vectors based on Histogram of Oriented Gradient (HOG). The Recurrent Neural Network (RNN) is then adapted to classify the s...

متن کامل

Aggregating Local Context for Accurate Scene Text Detection

Scene text reading continues to be of interest for many reasons including applications for the visually impaired and automatic image indexing systems. Here we propose a novel end-to-end scene text detection algorithm. First, for identifying text regions we design a novel Convolutional Neural Network (CNN) architecture that aggregates local surrounding information for cascaded, fast and accurate...

متن کامل

R2CNN: Rotational Region CNN for Orientation Robust Scene Text Detection

In this paper, we propose a novel method called Rotational Region CNN (R2CNN) for detecting arbitrary-oriented texts in natural scene images. The framework is based on Faster R-CNN [1] architecture. First, we use the Region Proposal Network (RPN) to generate axis-aligned bounding boxes that enclose the texts with different orientations. Second, for each axis-aligned text box proposed by RPN, we...

متن کامل

Scene text detection via extremal region based double threshold convolutional network classification

In this paper, we present a robust text detection approach in natural images which is based on region proposal mechanism. A powerful low-level detector named saliency enhanced-MSER extended from the widely-used MSER is proposed by incorporating saliency detection methods, which ensures a high recall rate. Given a natural image, character candidates are extracted from three channels in a percept...

متن کامل

Fused Text Segmentation Networks for Multi-oriented Scene Text Detection

In this paper, we introduce a novel end-end framework for multi-oriented scene text detection from an instanceaware segmentation perspective. We present Fused Text Segmentation Networks, which combine multi-level features during feature extracting as text instance may rely on finer feature expression compared to general objects. It detects and segments the text instance jointly and simultaneous...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Neurocomputing

سال: 2021

ISSN: ['0925-2312', '1872-8286']

DOI: https://doi.org/10.1016/j.neucom.2021.08.026